This GitHub repository, "agentic-ai-prompt-research" by Leonxlnx, contains a collection of prompts designed for use with agentic AI systems. The repository is organized into a series of markdown files, each representing a different prompt or prompt component.
Prompts cover a range of functionalities, including system prompts, simple modes, agent coordination, cyber risk instructions, and various skills like memory management, proactive behavior, and tool usage.
The prompts are likely intended for researchers and developers exploring and experimenting with the capabilities of autonomous AI agents. The collection aims to provide a resource for building more effective and robust agentic systems.
This repository focuses on the concept of an "agent" as a trained model, not just a framework or prompt chain. It emphasizes building a "harness" – the tools, knowledge, and interfaces that allow the model to function effectively in a specific domain. The core idea is that the model *is* the agent, and the engineer’s role is to create the environment it needs to succeed.
The content details a 12-session learning path, reverse-engineering the architecture of Claude Code to understand how to build robust and scalable agent harnesses. It highlights the importance of separating the agent (model) from the harness, and provides resources for extending this knowledge into practical applications.
Simon Willison explores "vibe coding" - building macOS apps with SwiftUI using large language models like Claude Opus 4.6 and GPT-5.4, without extensive coding knowledge. He successfully created two apps, Bandwidther (network bandwidth monitor) and Gpuer (GPU usage monitor), demonstrating the potential of this approach. The process involved minimal prompting and iterative development, leveraging the LLMs' capabilities for both code generation and feature suggestions.
While acknowledging the need for caution regarding the apps' accuracy, Willison highlights the efficiency and accessibility of building macOS applications in this manner.
This handbook provides a comprehensive introduction to Claude Code, Anthropic's AI-powered software development agent. It details how Claude Code differs from traditional autocomplete tools, functioning as an agent that reads, reasons about, and modifies codebases with user direction. The guide covers installation, initial setup, advanced workflows, integrations, and autonomous loops. It's aimed at developers, founders, and anyone seeking to leverage AI in software creation, emphasizing building real applications, accelerating feature development, and maintaining codebases efficiently. The handbook also highlights the importance of prompt discipline, planning, and understanding the underlying model to maximize Claude Code's capabilities.
CLI-Anything bridges the gap between AI agents and the world's software by making any software agent-ready. It's a universal interface for both humans and AI, offering a structured, lightweight, and self-describing approach. The project automates the creation of CLIs for applications like GIMP, Blender, and LibreOffice through a 7-phase pipeline – analyzing code, designing command groups, implementing the CLI, planning tests, writing tests, documenting, and publishing. It supports multiple platforms including Claude Code, OpenClaw, and Codex, with a focus on authentic software integration and production-grade testing.
This article advocates for wider adoption of Claude Code, an AI tool from Anthropic designed to write, edit, and fix code. Initially an internal tool for Anthropic developers, it's now publicly available as a command-line tool that operates within your terminal. It can understand natural language instructions to modify codebases, and even assists with non-programming tasks like file organization and research. While the terminal interface can be intimidating, the author suggests using it within an IDE or utilizing the Claude Desktop app's integrated Cowork interface, highlighting its potential for both developers and non-developers.
An account of how a developer, Alexey Grigorev, accidentally deleted 2.5 years of data from his AI Shipping Labs and DataTalks.Club websites using Claude Code and Terraform. Grigorev intended to migrate his website to AWS, but a missing state file and subsequent actions by Claude Code led to a complete wipe of the production setup, including the database and snapshots. The data was ultimately restored with help from Amazon Business support. The article highlights the importance of backups, careful permissions management, and manual review of potentially destructive actions performed by AI agents.
This article presents findings from a survey of over 900 software engineers regarding their use of AI tools. Key findings include the dominance of Claude Code, the mainstream adoption of AI in software engineering (95% weekly usage), the increasing use of AI agents (especially among staff+ engineers), and the influence of company size on tool choice. The survey also reveals which tools engineers love, with Claude Code being particularly favored, and provides demographic information about the respondents. A longer, 35-page report with additional details is available for full subscribers.
NanoClaw, a new open-source agent platform, aims to address the security concerns surrounding platforms like OpenClaw by utilizing containers and a smaller codebase. The project, started by Gavriel Cohen with the help of Anthropic's Claude Code, focuses on isolation and auditability, allowing agents to operate within a contained environment with limited access to system data.
This article details the setup and initial testing of Goose, an open-source agent framework, paired with Ollama and the Qwen3-coder model, as a free alternative to Claude Code. It covers the installation process, initial performance observations, and a comparison to cloud-based solutions.